± σ A ± t A ˆB ± σ B ± t B. (1)
|
|
- Morris Osborne
- 5 years ago
- Views:
Transcription
1 Version Introduction Combining measurements which have theoretical uncertainties is a delicate matter. Indeed, we are often in the position in which sufficient information is not provided to form a completely principled combination. However, we have developed a procedure which tends to avoid the worst pitfalls. We describe this procedure here. Suppose we are given two measurements, with results expressed in the form:  ± σ A ± t A ˆB ± σ B ± t B. () Assume that  has been sampled from a probability distribution of the form p A (Â; Ā, σ A), where Ā is the mean of the distribution and σ A is the standard deviation. We make the corresponding assumption for ˆB. The t A and t B uncertainties in Eq. are the theoretical uncertainties. We may not need to know exactly what that means here, except that the same meaning should hold for both t A and t B. We suppose that both  and ˆB are measurements of the same quantity of physical interest, though possibly with quite different approaches. The question is: How do we combine our two measurements? Let the physical quantity we are trying to learn about be denoted θ. Given the two results A and B, we wish to form an estimator, ˆV for V, with statistical and theoretical uncertainties expressed separately in the form: ˆV ± σ V ± t V. () The quantities ˆV, σ V,andt V are to be computed in terms of Â, ˆB,σ A,σ B,t A, and t B. Forming the Weighted Average In the absence of theoretical uncertainties, we would normally combine our measurements according to the weighted average: ˆθ =  σ A + ˆB σ B + σ σ A B ± σ A + σ B. (3) For simplicity, we are assuming here that there is no correlation between the measurements. In general,  and ˆB will be biased estimators for V : Ā = θ + b A B = θ + b B, (4)
2 where b A and b B are the biases. We adopt the point of view that the theoretical uncertainties t A and t B are estimates related to the possible magnitudes of these biases. That is, t A b A t B b B. (5) We wish to have t V represent a similar notion. Without yet specifying the weights, assume that we continue to form ˆV as a weighted average of  and ˆB: ˆθ = w A + w B ˆB, (6) w A + w B where w A and w B are the non-negative weights. The statistical error on the weighted average is computed according to simple error propagation on the individual statistical errors: The bias for ˆθ is: σ = w A σ A + w B σ B (w A + w B ). (7) b = ˆθ θ = w Ab A + w B b B. (8) w A + w B If the theoretical uncertainties are regarded as estimates of the biases, then the theoretical uncertainty should be evaluated with the same weighting: t = w At A + w B t B, (9) w A + w B It may be noted that this posesses desirable behavior in the limit where the theoretical uncertainties are identical (completely correlated) between the two measurements: The theoretical uncertainty on ˆV is in this case the same as t A = t B ; no reduction is attained by having multiple measurements. However, it is not quite true that the theoretical uncertainties are being regarded as estimates of bias. The t A and t B provide only estimates for the magnitudes, not the signs, of the biases. Eq. 9 holds when the biases are of the same sign. If the biases are opposite sign, then we obtain t = w At A w B t B. w A + w B (0) Thus, our formula 9 breaks down in some cases. For example, suppose the theoretical uncertainties are completely anticorrelated. In the case of equal weights, the combined theoretical uncertainty should be zero, because the two uncertainties are exactly canceled in the combined result. Only a statistical uncertainty remains.
3 Unfortunately, we don t always know whether the biases are expected to have the same sign or opposite sign. As a default, we adopt the procedure of Eq. 9. In the case of similar measurements, we suspect that the sign of the bias will often have the same sign, in which case we make the right choice. In the case of quite different measurements, such as inclusive and exclusive measurements of V ub, there is no particular reason to favor either relative sign; we simply don t know. The adopted procedure has the property that it errs on the side of conservatism we will sometimes overestimate the theoretical uncertainty on the combined result. There is still a further issue. The results of the measurements themselves canprovideinformationonwhatthetheoretical uncertainty could be. Consider two measurements with negligible statistical uncertainty. Then the difference between the two measurements is the difference between the biases. If the measurements are far apart, on the scale of the theoretical uncertainties, then this is evidence that the theoretical uncertainties are of opposite sign. We make no attempt to incorporate this information, again erring on the conservative side. We turn to the question of choice of weights w A and w B. In the limit of negligible theoretical uncertainties we want to have w A = σa () w B = σb. () Using these as the weights in the presence of theoretical uncertanties can lead to undesirable behavior. For example, suppose t A t B and σ A σ B. The central value computed with only the statistical weights ignores the theoretical uncertainty. A measurement with small theoretical uncertainty may be given little weight compared to a measurement with very large theoretical uncertainty. While not wrong, this does not make optimal use of the available information. We may invent a weighting scheme which incorporates both the statistical and theoretical uncertainties, for example combining them in quadrature: w A = w B = σa + t A σb +. (3) t B Any such scheme can lead to unattractive dependence on the way measurements may be associatively combined. We nevertheless adopt this procedure, with the understanding that it is best to go back to the original measureaments when combining results, rather than making successive combinations. 3
4 3 Inconsistent Inputs It may happen that our measurements are far enough apart that they appear inconsistent in terms of the quoted uncertainties. Our primary goal in this analysis is to test consistency between available data and the standard model, including whatever theoretical uncertainties exist in the comparison. We prefer to avoid making erroneous claims of inconsistency, even at the cost of some statistical power. Thus, we presume that when two measurements of what is assumed to be the same quantity appear inconsistent, something is wrong with the measurement or with the thoeretical uncertainties in the computation. If we have no good way to determine in detail where the fault lies, we adopt a method similar to that used by the Particle Data Group (PDG) to enlarge the stated uncertainties. Given our two measurements as discussed above, we define the quantity: χ w A (Â ˆθ) + w B ( ˆB ˆθ). (4) In the limit of purely statistical and normal errors, this quantity is distributed according to a chi-square with one degree of freedom. In the more general situation here, we don t know the detailed properties, but we nonetheless use it as a measure of the consistency of the results, in the belief that the procedure we adopt will still tend to err toward conservatism. If χ, the measurements are deemed consistent. On the other hand, if χ >, we call the measurements inconsistent, and apply a scale factor to the errors in order to obtain χ =. We take the point of view that we don t know which measurement (or both) is flawed, or whether the problem is with the statistical or theoretical error evaluation. If we did have such relevant information, we could use that in a more informed procedure. Thus, we scale all of the errors (σ A, σ B, t A, t B )byafactor: S = χ. (5) This scaling does not change the central value of the averaged result, but does scale the statistical and theoretical uncertainties by the same factor. 4 Relative Errors We often are faced with the situation in which the uncertainties are relative, rather than absolute. In this case, the model in which θ is a location parameter of a Gaussian distribution breaks down. However, it may be a reasonable approximation to continue to think in terms this model, with some modification to mitigate bias. We also continue to work in the context of a least-squares minimization, although it might be interesting to investigate a maximum likelihood approach. I believe that the approach suggested here is consistent with the proposal Bob Kowalewski is making for HFAG averages. 4
5 Thus, suppose we have additional experimental uncertainties s A and s B, which scale with θ: s A = r A θ, s B = r B θ. (6) If s k is what we are given, we infer the proportionality constants according to r A = s A /Â and r B = s B / ˆB. The weights that are given in Eqn. 3 are modified to incorporate this new source of uncertainty according to: w A = σa +(r ˆθ) A + t A w B = σb +(r ˆθ). (7) B + t B Note that, as we don t know θ, weuseˆθ instead. This means that the averaging process is now iterative, until convergence to a particular value of ˆθ is obtained. Likewise, there may be a theoretical uncertainty which scales with θ, and we may treat this similarly. Thus, suppose that, for example, t A = t aa t ra, where t aa is an absolute uncertainty, and t ra = ρ A θ. We simply replace θ by ˆθ and substitute this expression wherever t A appears, e.g., in Eqn. 7. That is: w A = σa +(r ˆθ) A + t aa +. (8) ρ A ˆθ 5 Summary of Algorithm We summarize the proposed algorithm: {x i i =,,...,n} with error matrix Suppose we have n measurements M ij (x i x i )(x j x j ), (9) and mean values x i = θ + b i. (0) Note that, in the non-correlated case, M ij = σi δ ij, or including relative uncertainties, M ij = δ ij (σi + r i x i ). The parameter we are trying to learn about is θ, andtheb i is the bias that is being estimated with theoretical uncertainties t i. The present notion of the weighted average is that we find a θ which minimizes: χ = (x i θ) W ij (x j θ). () i,j This is based on the premise that we don t actually know what the biases are, and we do the minimization with zero bias in the (x θ) dependence. The 5
6 possible size of bias is taken into account in the weighting, giving more weight to those measurements in which the size of the bias is likely to be smaller. The weight matrix W in principle could be taken to be: That is, W is an estimate for (W ) ij = M ij + t i t j. () (x i θ)(x j θ) = M ij + b i b j. (3) However, we don t assume that we know the relative signs of b i and b j. Hence, the off-diagonal t i t j term in Eqn. could just as likely enter with a minus sign. We therefore use the weight matrix: (W ) ij = M ij + t i δ ij. (4) If we do know the relative signs of the biases, for example because the theoretical uncertainties are correlated, then the off-diagonal terms in Eqn. should be included, with the appropriate sign. Setting dχ /dθ θ=ˆθ = 0 gives the central value ( best estimate): i,j ˆθ = W ijx j i,j W. (5) ij The statistical uncertainty is σ = i,j (WMW) ij i,j W ij (6) Note that this reduces to σ = i,j (M ) ij, (7) in the case of only statistical uncertainites. The theoretical uncertainty is i,j t = W ijt j i,j W, (8) ij where t j t aj +(ρ j ˆθ). (9) Finally, if χ >n, these error estimates are scaled by a factor: χ S = n, (30) where χ here is the value after the minimization. 6
7 A Comparison with treating theoretical uncertainties on same footing as statistical Another approach to the present problem is to simply treat the theoretical uncertainties as if they were statistical. This procedure gives the same estimator as above for θ. However, the results for statistical and theoretical uncertainties differ in general. Let σ be the estimated statistical uncertainty on the average for this approach, and let t be the estimated theoretical uncertainty. Also, let T ij be the covariance matrix for the theoretical uncertainties in this picture. Then the statistical and theoretical uncertainties on the average are given by: σ = t = n n Note that the weights are given, as before, by i,j,k M ijw jk i,j W ij, (3) i,j,k T ijw jk i,j W. (3) ij W ij =(M + T ) ij. (33) That is, the weights are the same as the treatment earlier, if the same assumptions about theoretical correlations are made in both places. The estimates for the statistical and theoretical uncertainties differ between the two methods. That is, in general, σ σ and t t. The statistical uncertainty σ is computed from the individual statistical uncertainties according to simple error propagation. The statistical uncertainty σ is evaluated by identifying a piece of the overall quadratic combination of statistical and theoretical uncertainties as statistical. The difference between t and t is that t is computed as a weighted average of the individual t s, while t is evaluated by identifying a piece of the overall quadratic combination of statistical and theoretical uncertainties as theoretical. The approach for t is based on the notion that the theoretical uncertainties are estimates of bias, but with a conservative treatment of any unknown correlations. The t approach may be appropriate if the theoretical uncertainties are given a probablistic interpretation. Let s consider some possible special cases. Suppose that all of the t i s are the same, equal to t, and suppose that the theory uncertainties are presumed to be uncorrelated. In this case, t = t (34) t = t / n. (35) Which is more reasonable? That depends on how we view the meaning of uncorrelated in our assumption, and on whether we assign a probabilistic 7
8 interpretation to the theoretical uncertainties. If we are supposing that the acutal theoretical uncertainties are somehow randomly distributed in sign and magnitude, then it is reasonable to expect that the result will become more reliable as more numbers are averaged. However, if we consider the theoretical uncertainties as estimates of bias, which could in fact all have the same sign, then the weighted linear average is plausible. It is at least a more conservative approach in the absence of real information on the correlations. Note that if the correlation in theoretical uncertainty is actually known, the weighted linear average will take that into account. For example, suppose there are just two measurements, with t = t. If the weights are the same (that is, we also have σ = σ )thent = 0. The other approach also gives( t =0. ) σ A different illustrative case is when t =0,t 0,andM = 0 0 σ.in this case, we find ( ˆθ = ˆθ x = σ + + x )/( t σ σ + + ) t σ, (36) σ σ = (σ + + /( t ) σ σ + + ) t σ, (37) t = t σ +t + σ +t σ ( σ = σ σ + t t = t σ + t, (38) )/( + /( σ + + t σ σ + + t σ ), (39) ). (40) To understand the difference better, consider the limit in which t σ,σ : ˆθ = ˆθ σ = x σ + + x x, (4) t σ = σ, (4) σ t = t σ + 0, (43) t σ = σ, (44) t = σ. (45) In this limit, both methods agree that the important information is in x.the first method assigns a statistical error corresponding to the statistical uncertainty of x, and a theoretical uncertainty of zero, reflecting the zero theoretical uncertainty on the x measurement. The second method, however, assigns equal 8
9 statistical and theoretical uncertainites to the average. Their sum in quadrature is a plausible expression of the total uncertainty, but the breakdown into theoretical and statistical components is not reasonable. Another limit we can take in this example is σ t σ, obtaining: ˆθ = ( ˆθ t ) = x + x σ x, (46) σ = σ (47) t = t, (48) σ = t, (49) t = t. (50) Similar observations may be made in this case as in the previous. 9
arxiv: v1 [physics.data-an] 3 Jun 2008
arxiv:0806.0530v [physics.data-an] 3 Jun 008 Averaging Results with Theory Uncertainties F. C. Porter Lauritsen Laboratory for High Energy Physics California Institute of Technology Pasadena, California
More informationPhysics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester
Physics 403 Propagation of Uncertainties Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals
More informationFitting a Straight Line to Data
Fitting a Straight Line to Data Thanks for your patience. Finally we ll take a shot at real data! The data set in question is baryonic Tully-Fisher data from http://astroweb.cwru.edu/sparc/btfr Lelli2016a.mrt,
More informationParameter Estimation and Fitting to Data
Parameter Estimation and Fitting to Data Parameter estimation Maximum likelihood Least squares Goodness-of-fit Examples Elton S. Smith, Jefferson Lab 1 Parameter estimation Properties of estimators 3 An
More informationGaussian Quiz. Preamble to The Humble Gaussian Distribution. David MacKay 1
Preamble to The Humble Gaussian Distribution. David MacKay Gaussian Quiz H y y y 3. Assuming that the variables y, y, y 3 in this belief network have a joint Gaussian distribution, which of the following
More informationPhysics 509: Propagating Systematic Uncertainties. Scott Oser Lecture #12
Physics 509: Propagating Systematic Uncertainties Scott Oser Lecture #1 1 Additive offset model Suppose we take N measurements from a distribution, and wish to estimate the true mean of the underlying
More informationModern Methods of Data Analysis - WS 07/08
Modern Methods of Data Analysis Lecture VIa (19.11.07) Contents: Uncertainties (II): Re: error propagation Correlated uncertainties Systematic uncertainties Re: Error Propagation (I) x = Vi,j and µi known
More informationStatistical Models with Uncertain Error Parameters (G. Cowan, arxiv: )
Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv:1809.05778) Workshop on Advanced Statistics for Physics Discovery aspd.stat.unipd.it Department of Statistical Sciences, University of
More informationEM Waves in Media. What happens when an EM wave travels through our model material?
EM Waves in Media We can model a material as made of atoms which have a charged electron bound to a nucleus by a spring. We model the nuclei as being fixed to a grid (because they are heavy, they don t
More informationMIXED MODELS THE GENERAL MIXED MODEL
MIXED MODELS This chapter introduces best linear unbiased prediction (BLUP), a general method for predicting random effects, while Chapter 27 is concerned with the estimation of variances by restricted
More informationChoosing among models
Eco 515 Fall 2014 Chris Sims Choosing among models September 18, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported
More informationStatistics and Data Analysis
Statistics and Data Analysis The Crash Course Physics 226, Fall 2013 "There are three kinds of lies: lies, damned lies, and statistics. Mark Twain, allegedly after Benjamin Disraeli Statistics and Data
More informationUnobserved heterogeneity in logistic regression
Unobserved heterogeneity in logistic regression Maarten L. Buis November 18, 2005 1 Introduction Regression is used to find the relation between various eplanatory variables and an eplained variable. For
More informationThe variable θ is called the parameter of the model, and the set Ω is called the parameter space.
Lecture 8 What is a statistical model? A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced the data. The variable θ is called
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationA8824: Statistics Notes David Weinberg, Astronomy 8824: Statistics Notes 6 Estimating Errors From Data
Where does the error bar go? Astronomy 8824: Statistics otes 6 Estimating Errors From Data Suppose you measure the average depression of flux in a quasar caused by absorption from the Lyman-alpha forest.
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 12: Frequentist properties of estimators (v4) Ramesh Johari ramesh.johari@stanford.edu 1 / 39 Frequentist inference 2 / 39 Thinking like a frequentist Suppose that for some
More informationLinear Regression Models
Linear Regression Models Model Description and Model Parameters Modelling is a central theme in these notes. The idea is to develop and continuously improve a library of predictive models for hazards,
More informationSo, what are special sciences? ones that are particularly dear to the author? ( Oh dear. I am touched. Psychology is just, so, well, special!
Jerry Fodor and his Special Sciences So, what are special sciences? ones that are particularly dear to the author? ( Oh dear. I am touched. Psychology is just, so, well, special! ) The use of special in
More informationPurposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions
Part 1: Probability Distributions Purposes of Data Analysis True Distributions or Relationships in the Earths System Probability Distribution Normal Distribution Student-t Distribution Chi Square Distribution
More informationParameter Uncertainties in IMFIT CASA Requirements Document Brian S. Mason (NRAO) March 5, 2014
Parameter Uncertainties in IMFIT CASA Requirements Document Brian S. Mason (NRAO) March 5, 014 The CASA task IMFIT fits elliptical Gaussians to image data. As reported by several users in CASA tickets
More informationFrequentist-Bayesian Model Comparisons: A Simple Example
Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal
More informationHypothesis testing. Chapter Formulating a hypothesis. 7.2 Testing if the hypothesis agrees with data
Chapter 7 Hypothesis testing 7.1 Formulating a hypothesis Up until now we have discussed how to define a measurement in terms of a central value, uncertainties, and units, as well as how to extend these
More informationHistograms were invented for human beings, so that we can visualize the observations as a density (estimated values of a pdf).
How to fit a histogram The Histogram Histograms were invented for human beings, so that we can visualize the observations as a density (estimated values of a pdf). The binning most convenient for visualization
More information4th IMPRS Astronomy Summer School Drawing Astrophysical Inferences from Data Sets
4th IMPRS Astronomy Summer School Drawing Astrophysical Inferences from Data Sets William H. Press The University of Texas at Austin Lecture 6 IMPRS Summer School 2009, Prof. William H. Press 1 Mixture
More informationPHYSTAT2003, SLAC, Stanford, California, September 8-11, 2003
Asymmetric Errors PHYSTAT003, SLAC, Stanford, California, September 8-11, 003 Roger Barlow Manchester University, UK and Stanford University, USA Errors quoted on results are often given in asymmetric
More informationSystematic uncertainties in statistical data analysis for particle physics. DESY Seminar Hamburg, 31 March, 2009
Systematic uncertainties in statistical data analysis for particle physics DESY Seminar Hamburg, 31 March, 2009 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan
More informationDS-GA 1002 Lecture notes 11 Fall Bayesian statistics
DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian
More informationSIO 221B, Rudnick adapted from Davis 1. 1 x lim. N x 2 n = 1 N. { x} 1 N. N x = 1 N. N x = 1 ( N N x ) x = 0 (3) = 1 x N 2
SIO B, Rudnick adapted from Davis VII. Sampling errors We do not have access to the true statistics, so we must compute sample statistics. By this we mean that the number of realizations we average over
More informationTreatment of Error in Experimental Measurements
in Experimental Measurements All measurements contain error. An experiment is truly incomplete without an evaluation of the amount of error in the results. In this course, you will learn to use some common
More informationBayesian model selection: methodology, computation and applications
Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program
More informationJourneys of an Accidental Statistician
Journeys of an Accidental Statistician A partially anecdotal account of A Unified Approach to the Classical Statistical Analysis of Small Signals, GJF and Robert D. Cousins, Phys. Rev. D 57, 3873 (1998)
More information6.867 Machine Learning
6.867 Machine Learning Problem set 1 Solutions Thursday, September 19 What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove.
More informationDS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling
DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling Due: Tuesday, May 10, 2016, at 6pm (Submit via NYU Classes) Instructions: Your answers to the questions below, including
More informationarxiv:hep-ex/ v1 2 Jun 2000
MPI H - V7-000 May 3, 000 Averaging Measurements with Hidden Correlations and Asymmetric Errors Michael Schmelling / MPI for Nuclear Physics Postfach 03980, D-6909 Heidelberg arxiv:hep-ex/0006004v Jun
More informationThe Multivariate Gaussian Distribution [DRAFT]
The Multivariate Gaussian Distribution DRAFT David S. Rosenberg Abstract This is a collection of a few key and standard results about multivariate Gaussian distributions. I have not included many proofs,
More informationEcon 836 Final Exam. 2 w N 2 u N 2. 2 v N
1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.
More informationVectors Part 1: Two Dimensions
Vectors Part 1: Two Dimensions Last modified: 20/02/2018 Links Scalars Vectors Definition Notation Polar Form Compass Directions Basic Vector Maths Multiply a Vector by a Scalar Unit Vectors Example Vectors
More informationCSC 578 Neural Networks and Deep Learning
CSC 578 Neural Networks and Deep Learning Fall 2018/19 3. Improving Neural Networks (Some figures adapted from NNDL book) 1 Various Approaches to Improve Neural Networks 1. Cost functions Quadratic Cross
More informationOverview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation
Overview Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation Probabilistic Interpretation: Linear Regression Assume output y is generated
More information1 Motivation for Instrumental Variable (IV) Regression
ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data
More informationMACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION
MACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION THOMAS MAILUND Machine learning means different things to different people, and there is no general agreed upon core set of algorithms that must be
More informationInterval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean
Interval estimation October 3, 2018 STAT 151 Class 7 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0
More informationOn the errors introduced by the naive Bayes independence assumption
On the errors introduced by the naive Bayes independence assumption Author Matthijs de Wachter 3671100 Utrecht University Master Thesis Artificial Intelligence Supervisor Dr. Silja Renooij Department of
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 10 Alternatives to Monte Carlo Computation Since about 1990, Markov chain Monte Carlo has been the dominant
More informationECONOMICS 210C / ECONOMICS 236A MONETARY HISTORY
Fall 018 University of California, Berkeley Christina Romer David Romer ECONOMICS 10C / ECONOMICS 36A MONETARY HISTORY A Little about GLS, Heteroskedasticity-Consistent Standard Errors, Clustering, and
More informationBAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS
BAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS Andrew A. Neath 1 and Joseph E. Cavanaugh 1 Department of Mathematics and Statistics, Southern Illinois University, Edwardsville, Illinois 606, USA
More informationProbabilistic Graphical Models
2016 Robert Nowak Probabilistic Graphical Models 1 Introduction We have focused mainly on linear models for signals, in particular the subspace model x = Uθ, where U is a n k matrix and θ R k is a vector
More informationPh 219/CS 219. Exercises Due: Friday 20 October 2006
1 Ph 219/CS 219 Exercises Due: Friday 20 October 2006 1.1 How far apart are two quantum states? Consider two quantum states described by density operators ρ and ρ in an N-dimensional Hilbert space, and
More informationMeasurement And Uncertainty
Measurement And Uncertainty Based on Guidelines for Evaluating and Expressing the Uncertainty of NIST Measurement Results, NIST Technical Note 1297, 1994 Edition PHYS 407 1 Measurement approximates or
More informationWooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares
Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares Many economic models involve endogeneity: that is, a theoretical relationship does not fit
More informationIV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors
IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors Laura Mayoral IAE, Barcelona GSE and University of Gothenburg Gothenburg, May 2015 Roadmap Deviations from the standard
More informationECON 4160, Spring term 2015 Lecture 7
ECON 4160, Spring term 2015 Lecture 7 Identification and estimation of SEMs (Part 1) Ragnar Nymoen Department of Economics 8 Oct 2015 1 / 55 HN Ch 15 References to Davidson and MacKinnon, Ch 8.1-8.5 Ch
More informationIntroductory Econometrics. Review of statistics (Part II: Inference)
Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16 Null and alternative hypotheses Usually, we have two competing
More informationStatistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That
Statistics Lecture 4 August 9, 2000 Frank Porter Caltech The plan for these lectures: 1. The Fundamentals; Point Estimation 2. Maximum Likelihood, Least Squares and All That 3. What is a Confidence Interval?
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationINTERPENETRATING SAMPLES
INTERPENETRATING SAMPLES Dr. A. C. Kulshreshtha UN Statistical Institute for Asia and Pacific Second RAP Regional Workshop on Building Training Resources for Improving Agricultural & Rural Statistics Sampling
More informationBayesian vs frequentist techniques for the analysis of binary outcome data
1 Bayesian vs frequentist techniques for the analysis of binary outcome data By M. Stapleton Abstract We compare Bayesian and frequentist techniques for analysing binary outcome data. Such data are commonly
More informationReview of Maximum Likelihood Estimators
Libby MacKinnon CSE 527 notes Lecture 7, October 7, 2007 MLE and EM Review of Maximum Likelihood Estimators MLE is one of many approaches to parameter estimation. The likelihood of independent observations
More informationAn Empirical-Bayes Score for Discrete Bayesian Networks
An Empirical-Bayes Score for Discrete Bayesian Networks scutari@stats.ox.ac.uk Department of Statistics September 8, 2016 Bayesian Network Structure Learning Learning a BN B = (G, Θ) from a data set D
More information32. STATISTICS. 32. Statistics 1
32. STATISTICS 32. Statistics 1 Revised September 2007 by G. Cowan (RHUL). This chapter gives an overview of statistical methods used in High Energy Physics. In statistics, we are interested in using a
More information1 What does the random effect η mean?
Some thoughts on Hanks et al, Environmetrics, 2015, pp. 243-254. Jim Hodges Division of Biostatistics, University of Minnesota, Minneapolis, Minnesota USA 55414 email: hodge003@umn.edu October 13, 2015
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationUncertainty and Rules
Uncertainty and Rules We have already seen that expert systems can operate within the realm of uncertainty. There are several sources of uncertainty in rules: Uncertainty related to individual rules Uncertainty
More informationIntroduction to Bayesian Learning. Machine Learning Fall 2018
Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability
More informationDetection and quantification capabilities
18.4.3.7 Detection and quantification capabilities Among the most important Performance Characteristics of the Chemical Measurement Process (CMP) are those that can serve as measures of the underlying
More informationPhysics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester
Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability
More informationBayesian inference: what it means and why we care
Bayesian inference: what it means and why we care Robin J. Ryder Centre de Recherche en Mathématiques de la Décision Université Paris-Dauphine 6 November 2017 Mathematical Coffees Robin Ryder (Dauphine)
More informationClass President: A Network Approach to Popularity. Due July 18, 2014
Class President: A Network Approach to Popularity Due July 8, 24 Instructions. Due Fri, July 8 at :59 PM 2. Work in groups of up to 3 3. Type up the report, and submit as a pdf on D2L 4. Attach the code
More informationIntroduction: MLE, MAP, Bayesian reasoning (28/8/13)
STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this
More information37-6 Watching the electrons (matter waves)
37-6 Watching the electrons (matter waves) 1 testing our proposition: the electrons go either through hole 1 or hole 2 add a very strong light source behind walls between two holes, electrons will scatter
More informationStrati cation in Multivariate Modeling
Strati cation in Multivariate Modeling Tihomir Asparouhov Muthen & Muthen Mplus Web Notes: No. 9 Version 2, December 16, 2004 1 The author is thankful to Bengt Muthen for his guidance, to Linda Muthen
More informationQuantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing
Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu October
More informationMeasurement Error PHYS Introduction
PHYS 1301 Measurement Error Introduction We have confidence that a particular physics theory is telling us something interesting about the physical universe because we are able to test quantitatively its
More informationWeek 3: Linear Regression
Week 3: Linear Regression Instructor: Sergey Levine Recap In the previous lecture we saw how linear regression can solve the following problem: given a dataset D = {(x, y ),..., (x N, y N )}, learn to
More informationDiscrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10
EECS 70 Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 Introduction to Basic Discrete Probability In the last note we considered the probabilistic experiment where we flipped
More informationUnified approach to the classical statistical analysis of small signals
PHYSICAL REVIEW D VOLUME 57, NUMBER 7 1 APRIL 1998 Unified approach to the classical statistical analysis of small signals Gary J. Feldman * Department of Physics, Harvard University, Cambridge, Massachusetts
More informationUC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics. EECS 281A / STAT 241A Statistical Learning Theory
UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics EECS 281A / STAT 241A Statistical Learning Theory Solutions to Problem Set 2 Fall 2011 Issued: Wednesday,
More informationDEMBSKI S SPECIFIED COMPLEXITY: A SIMPLE ERROR IN ARITHMETIC
DEMBSKI S SPECIFIED COMPLEXITY: A SIMPLE ERROR IN ARITHMETIC HOWARD A. LANDMAN Abstract. We show that the derivation of Dembski s formula for specified complexity contains a simple but enormous error,
More informationMeasurement Error PHYS Introduction
PHYS 1301 Measurement Error Introduction We have confidence that a particular physics theory is telling us something interesting about the physical universe because we are able to test quantitatively its
More informationFrank Porter February 26, 2013
116 Frank Porter February 26, 2013 Chapter 6 Hypothesis Tests Often, we want to address questions such as whether the possible observation of a new effect is really significant, or merely a chance fluctuation.
More informationAdvanced Statistical Methods. Lecture 6
Advanced Statistical Methods Lecture 6 Convergence distribution of M.-H. MCMC We denote the PDF estimated by the MCMC as. It has the property Convergence distribution After some time, the distribution
More informationContinuum Probability and Sets of Measure Zero
Chapter 3 Continuum Probability and Sets of Measure Zero In this chapter, we provide a motivation for using measure theory as a foundation for probability. It uses the example of random coin tossing to
More informationLecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012
Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed
More informationAges of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008
Ages of stellar populations from color-magnitude diagrams Paul Baines Department of Statistics Harvard University September 30, 2008 Context & Example Welcome! Today we will look at using hierarchical
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More informationECON 4160, Autumn term Lecture 1
ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least
More informationAn improved procedure for combining Type A and Type B components of measurement uncertainty
Int. J. Metrol. Qual. Eng. 4, 55 62 (2013) c EDP Sciences 2013 DOI: 10.1051/ijmqe/2012038 An improved procedure for combining Type A and Type B components of measurement uncertainty R. Willink Received:
More informationFor more information about how to cite these materials visit
Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/
More informationLecture 24: Weighted and Generalized Least Squares
Lecture 24: Weighted and Generalized Least Squares 1 Weighted Least Squares When we use ordinary least squares to estimate linear regression, we minimize the mean squared error: MSE(b) = 1 n (Y i X i β)
More informationQuadrature for the Finite Free Convolution
Spectral Graph Theory Lecture 23 Quadrature for the Finite Free Convolution Daniel A. Spielman November 30, 205 Disclaimer These notes are not necessarily an accurate representation of what happened in
More informationHomework #2 Due Monday, April 18, 2012
12.540 Homework #2 Due Monday, April 18, 2012 Matlab solution codes are given in HW02_2012.m This code uses cells and contains the solutions to all the questions. Question 1: Non-linear estimation problem
More informationBrandon C. Kelly (Harvard Smithsonian Center for Astrophysics)
Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Probability quantifies randomness and uncertainty How do I estimate the normalization and logarithmic slope of a X ray continuum, assuming
More informationSimple groups and the classification of finite groups
Simple groups and the classification of finite groups 1 Finite groups of small order How can we describe all finite groups? Before we address this question, let s write down a list of all the finite groups
More informationA Bayesian Treatment of Linear Gaussian Regression
A Bayesian Treatment of Linear Gaussian Regression Frank Wood December 3, 2009 Bayesian Approach to Classical Linear Regression In classical linear regression we have the following model y β, σ 2, X N(Xβ,
More information32. STATISTICS. 32. Statistics 1
32. STATISTICS 32. Statistics 1 Revised April 1998 by F. James (CERN); February 2000 by R. Cousins (UCLA); October 2001, October 2003, and August 2005 by G. Cowan (RHUL). This chapter gives an overview
More informationChapter 7: Hypothesis testing
Chapter 7: Hypothesis testing Hypothesis testing is typically done based on the cumulative hazard function. Here we ll use the Nelson-Aalen estimate of the cumulative hazard. The survival function is used
More informationHandout 1 Systems of linear equations Gaussian elimination
06-21254 Mathematical Techniques for Computer Science The University of Birmingham Autumn Semester 2017 School of Computer Science c Achim Jung September 27, 2017 Handout 1 Systems of linear equations
More informationTesting Restrictions and Comparing Models
Econ. 513, Time Series Econometrics Fall 00 Chris Sims Testing Restrictions and Comparing Models 1. THE PROBLEM We consider here the problem of comparing two parametric models for the data X, defined by
More information